Skip to content

Conversation

@MAXNORM8650
Copy link

@MAXNORM8650 MAXNORM8650 commented Aug 13, 2025

  • Add LlamaCppProvider with background server management
  • Support for TinyLlama, Gemma-3-4B, and SmolLM3 models
  • Automatic model downloading from Hugging Face
  • Silent server operation with clean command output
  • Add sample configurations for different models
  • Update README with LlamaCpp setup instructions
  • Server runs in background until manually stopped

Supports models in simple config and can be extended with all smol models:

  • tinyllama-1.1b (fast, basic responses)
  • gemma-3-4b (balanced quality/speed)
  • smollm3-3b (small, efficient)

Usage: Set LLAMA_DIR env var and use config commands to switch models

- Add LlamaCppProvider with background server management
- Support for TinyLlama, Gemma-3-4B, and SmolLM3 models
- Automatic model downloading from Hugging Face
- Silent server operation with clean command output
- Add sample configurations for different models
- Update README with LlamaCpp setup instructions
- Server runs in background until manually stopped

Supports models:
- tinyllama-1.1b (fast, basic responses)
- gemma-3-4b (balanced quality/speed)
- smollm3-3b (small, efficient)

Usage: Set LLAMA_DIR env var and use config commands to switch models
@samheutmaker
Copy link
Contributor

@MAXNORM8650 Can you resolve conflicts?

@MAXNORM8650
Copy link
Author

I have resolved all the conflicts except new features related eg. type ProviderType = "OpenAI" | "Custom" | "Claude" | "Gemini" | "GitHub" | "LlamaCpp"; vs type ProviderType = "OpenAI" | "Custom" | "Claude" | "Gemini" | "GitHub";.

@MAXNORM8650
Copy link
Author

Can you please look into these two conflicts, which is related to the llama cpp new features? I am not sure how to resolve both conflicts, as they are new features related to Llama cpp.

@samheutmaker
Copy link
Contributor

Please fix the conflicts.

@UniquePratham
Copy link

@samheutmaker Hey can I look into into to resolve the conflicts

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants